In-Memory Data Analytics on Coupled CPU-GPU Architectures
نویسندگان
چکیده
In the big data era, in-memory data analytics is an effective means of achieving high performance data processing and realizing the value of data in a timely manner. Efforts in this direction have been spent on various aspects, including in-memory algorithmic designs and system optimizations. In this paper, we propose to develop the next-generation in-memory relational database processing techniques on coupled CPU-GPU architectures. Particularly, we demonstrate novel design and implementations of query processing paradigms to utilize the strengths of coupled CPU-GPU architectures such as shared main memory and cache hierarchy. We propose a fine-grained method to distribute workload onto available processors, since the CPU and the GPU share the same main memory space. Besides, we propose an in-cache paradigm for query processing to take advantage of shared cache hierarchy to overcome memory stalls of query processing. Our experimental results demonstrate that 1) the proposed finegrained and in-cache query processing significantly improve the performance of in-memory databases, and 2) such coupled architectures are more energy efficient in query processing compared with other discrete systems.
منابع مشابه
Revisiting Co-Processing for Hash Joins on the Coupled CPU-GPU Architecture
Query co-processing on graphics processors (GPUs) has become an effective means to improve the performance of main memory databases. However, the relatively low bandwidth and high latency of the PCI-e bus are usually bottleneck issues for co-processing. Recently, coupled CPU-GPU architectures have received a lot of attention, e.g. AMD APUs with the CPU and the GPU integrated into a single chip....
متن کاملIn-Cache Query Co-Processing on Coupled CPU-GPU Architectures
Recently, there have been some emerging processor designs that the CPU and the GPU (Graphics Processing Unit) are integrated in a single chip and share Last Level Cache (LLC). However, the main memory bandwidth of such coupled CPU-GPU architectures can be much lower than that of a discrete GPU. As a result, current GPU query coprocessing paradigms can severely suffer from memory stalls. In this...
متن کاملAdvances in GPU Research and Practice
Tesla NVIDIA computing accelerators are currently based on Kepler and Maxwell architectures. The recent versions of Compute Unified Device Architecture (CUDA), such as CUDA 7.0, coupled with the Kepler and Maxwell architectures facilitate the dynamic use of GPUs. Moreover, data transfers can now happen via high-speed network directly from any GPU memory to any other GPU memory in any other clus...
متن کاملGPU-Accelerated Large Scale Analytics
In this paper, we report our research on using GPUs as accelerators for Business Intelligence(BI) analytics. We are particularly interested in analytics on very large data sets, which are common in today's real world BI applications. While many published works have shown that GPUs can be used to accelerate various general purpose applications with respectable performance gains, few attempts hav...
متن کاملMemory Compression Coordinated and Optimized Prefetching in GPU Architectures
Traditionally, GPU architectures have been primarily focused on throughput and latency hiding. However, as the computational power of GPUs continues to scale with Moore’s law, an increasing number of applications are becoming limited by memory bandwidth [1]. Also, data locality and reuse are becoming increasingly important with power-limited technology scaling. The energy spent on off-chip memo...
متن کامل